Predicting Human Eye Fixations via an LSTM-based Saliency Attentive Model

نویسندگان

  • Marcella Cornia
  • Lorenzo Baraldi
  • Giuseppe Serra
  • Rita Cucchiara
چکیده

Data-driven saliency has recently gained a lot of attention thanks to the use of Convolutional Neural Networks. In this paper we go beyond the standard approach to saliency prediction, in which gaze maps are computed with a feedforward network, and we present a novel Saliency Attentive Model which can predict accurate saliency maps by incorporating attentive mechanisms. Our solution is composed of a Convolutional LSTM, that iteratively focuses on the most salient regions of the input, and a Residual Architecture designed to preserve spatial resolution. Additionally, to tackle the center bias present in human eye fixations, our model incorporates prior maps generated by learned Gaussian functions. We show, through an extensive evaluation, that the proposed architecture overcomes the current state of the art on three public saliency prediction datasets: SALICON, MIT300 and CAT2000. We further study the contribution of each key components to demonstrate their robustness on different scenarios.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Supersaliency: Predicting Smooth Pursuit-Based Attention with Slicing CNNs Improves Fixation Prediction for Naturalistic Videos

Predicting attention is a popular topic at the intersection of human and computer vision, but video saliency prediction has only recently begun to benefit from deep learning-based approaches. Even though most of the available video-based saliency data sets and models claim to target human observers’ fixations, they fail to differentiate them from smooth pursuit (SP), a major eye movement type t...

متن کامل

Predicting Video Saliency with Object-to-Motion CNN and Two-layer Convolutional LSTM

Over the past few years, deep neural networks (DNNs) have exhibited great success in predicting the saliency of images. However, there are few works that apply DNNs to predict the saliency of generic videos. In this paper, we propose a novel DNN-based video saliency prediction method. Specifically, we establish a large-scale eye-tracking database of videos (LEDOV), which provides sufficient dat...

متن کامل

A Benchmark of Computational Models of Saliency to Predict Human Fixations

Many computational models of visual attention have been created from a wide variety of different approaches to predict where people look in images. Each model is usually introduced by demonstrating performances on new images, and it is hard to make immediate comparisons between models. To alleviate this problem, we propose a benchmark data set containing 300 natural images with eye tracking dat...

متن کامل

Visual Attention for Rendered 3D Shapes

Understanding the attentional behavior of the human visual system when visualizing a rendered 3D shape is of great importance for many computer graphics applications. Eye tracking remains the only solution to explore this complex cognitive mechanism. Unfortunately, despite the large number of studies dedicated to images and videos, only a few eye tracking experiments have been conducted using 3...

متن کامل

Augmented saliency model using automatic 3D head pose detection and learned gaze following in natural scenes

Previous studies have shown that gaze direction of actors in a scene influences eye movements of passive observers during free-viewing (Castelhano, Wieth, & Henderson, 2007; Borji, Parks, & Itti, 2014). However, no computational model has been proposed to combine bottom-up saliency with actor's head pose and gaze direction for predicting where observers look. Here, we first learn probability ma...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:
  • CoRR

دوره abs/1611.09571  شماره 

صفحات  -

تاریخ انتشار 2016